MASTER THESIS by Paul Fink Ensemble methods for classification trees under imprecise probabilities

نویسنده

  • Paul Fink
چکیده

In this master thesis some properties of bags of imprecise classification trees, as introduced in Abellán and Masegosa (2010), are analysed. In the beginning the statistical background of imprecise classification trees is outlined – starting with an overview on measuring uncertainty within the concept of Dempster–Shafer theory is presented, followed by a discussion of its application in a tree–growing–algorithm, which employs the so–called Imprecise Dirichlet Model in the splitting process. The motivation of so–called ensemble methods is to reduce the instability of a single classification tree, increasing its predictive accuracy, but at cost of structural interpretability. A description of the well known ensemble methods, such as bagging, random forests and boosting, is given along with two approaches, generating the ensemble by allowing more than one splitting variable within a node. In the next step a bag of imprecise classification trees is generated; following, its sensitivity in relation to different ensemble aggregation/fusion rules (majority voting, disjunction and average rule), the external stopping criterion of minimal observations within a node and the main parameter of the Imprecise Dirichlet Model is analysed by a simulation study. The results of the simulation indicate that the commonly applied majority voting rule is also a fair choice for imprecise classification ensembles. Regarding the external stopping criterion the simulation indicates that such may be neglected, while the parameter does highly affect the predictive accuracy, favouring lower values of it.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Variable Selection Bias in Classification Trees Based on Imprecise Probabilities

Classification trees are a popular statistical tool with multiple applications. Recent advancements of traditional classification trees, such as the approach of classification trees based on imprecise probabilities by Abellán and Moral (2005), effectively address their tendency to overfitting. However, another flaw inherent in traditional classification trees is not eliminated by the imprecise ...

متن کامل

Entropy based classification trees

One method for building classification trees is to choose split variables by maximising expected entropy. This can be extended through the application of imprecise probability by replacing instances of expected entropy with the maximum possible expected entropy over credal sets of probability distributions. Such methods may not take full advantage of the opportunities offered by imprecise proba...

متن کامل

Variable Selection in Classification Trees Based on Imprecise Probabilities

Classification trees are a popular statistical tool with multiple applications. Recent advancements of traditional classification trees, such as the approach of classification trees based on imprecise probabilities by Abellán and Moral (2004), effectively address their tendency to overfitting. However, another flaw inherent in traditional classification trees is not eliminated by the imprecise ...

متن کامل

Statistical Modelling in Surveys without Neglecting The Undecided: Multinomial Logistic Regression Models and Imprecise Classification Trees under Ontic Data Imprecision

In surveys, and most notably in election polls, undecided participants frequently constitute subgroups of their own with specific individual characteristics. While traditional survey methods and corresponding statistical models are inherently damned to neglect this valuable information, an ontic random set view provides us with the full power of the whole statistical modelling framework. We ela...

متن کامل

Multinomial Nonparametric Predictive Inference: Selection, Classification and Subcategory Data

In probability and statistics, uncertainty is usually quantified using single-valued probabilities satisfying Kolmogorov’s axioms. Generalisation of classical probability theory leads to various less restrictive representations of uncertainty which are collectively referred to as imprecise probability. Several approaches to statistical inference using imprecise probability have been suggested, ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012